A Tractable POMDP for a Class of Sequencing Problems

نویسندگان

  • Paat Rusmevichientong
  • Benjamin Van Roy
چکیده

We consider a partially observable Markov decision problem (POMDP) that models a class of sequencing problems. Although POMDPs are typically intractable, our formulation admits tractable solution. Instead of maintaining a value function over a high-dimensional set of belief states, we reduce the state space to one of smaller dimension, in which grid-based dynamic programming techniques are effective. We develop an error bound for the resulting approximation, and discuss an application of the model to a problem in targeted advertising. Subject classifications: Dynamic programming: partially observable Markov decision problem. Decision analysis: sequential. Marketing: targeted advertising.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A POMDP Framework to Find Optimal Inspection and Maintenance Policies via Availability and Profit Maximization for Manufacturing Systems

Maintenance can be the factor of either increasing or decreasing system's availability, so it is valuable work to evaluate a maintenance policy from cost and availability point of view, simultaneously and according to decision maker's priorities. This study proposes a Partially Observable Markov Decision Process (POMDP) framework for a partially observable and stochastically deteriorating syste...

متن کامل

A Single Machine Sequencing Problem with Idle Insert: Simulated Annealing and Branch-and-Bound Methods

  In this paper, a single machine sequencing problem is considered in order to find the sequence of jobs minimizing the sum of the maximum earliness and tardiness with idle times (n/1/I/ETmax). Due to the time complexity function, this sequencing problem belongs to a class of NP-hard ones. Thus, a special design of a simulated annealing (SA) method is applied to solve such a hard problem. To co...

متن کامل

Tractable POMDP Planning Algorithms for Optimal Teaching in “SPAIS”

In this paper, we develop a system for teaching the task of sorting a set of virtual coins. Teaching is a challenging domain for AI systems because three problems must be solved at once: a teacher must simultaneously infer both social variables (attention, boredom, confusion, expertise, aptitude), as well as physical ones (task progress, objects being used, current activity), and finally she mu...

متن کامل

Tractable Planning under Uncertainty: Exploiting Structure

T HE problem of planning under uncertainty has received significant attention in the scientific community over the past few years. It is now well-recognized that considering uncertainty during planning and decision-making is imperative to the design of robust computer systems. This is particularly crucial in robotics, where the ability to interact effectively with real-world environments is a p...

متن کامل

Towards Solving Large-Scale POMDP Problems Via Spatio-Temporal Belief State Clustering

Markov decision process (MDP) is commonly used to model a stochastic environment for supporting optimal decision making. However, solving a large-scale MDP problem under the partially observable condition (also called POMDP) is known to be computationally intractable. Belief compression by reducing belief state dimension has recently been shown to be an effective way for making the problem trac...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1301.2308  شماره 

صفحات  -

تاریخ انتشار 2001